Goto

Collaborating Authors

 pew research center


Artificial Authority: From Machine Minds to Political Alignments. An Experimental Analysis of Democratic and Autocratic Biases in Large-Language Models

Ożegalska-Łukasik, Natalia, Łukasik, Szymon

arXiv.org Artificial Intelligence

Political beliefs vary significantly across different countries, reflecting distinct historical, cultural, and institutional contexts. These ideologies, ranging from liberal democracies to rigid autocracies, influence human societies, as well as the digital systems that are constructed within those societies. The advent of generative artificial intelligence, particularly Large Language Models (LLMs), introduces new agents in the political space-agents trained on massive corpora that replicate and proliferate socio-political assumptions. This paper analyses whether LLMs display propensities consistent with democratic or autocratic world-views. We validate this insight through experimental tests in which we experiment with the leading LLMs developed across disparate political contexts, using several existing psychometric and political orientation measures. The analysis is based on both numerical scoring and qualitative analysis of the models' responses. Findings indicate high model-to-model variability and a strong association with the political culture of the country in which the model was developed. These findings highlight the need for more detailed examination of the socio-political dimensions embedded within AI systems.


PrimeX: A Dataset of Worldview, Opinion, and Explanation

Koncel-Kedziorski, Rik, Joshi, Brihi, Paek, Tim

arXiv.org Artificial Intelligence

As the adoption of language models advances, so does the need to better represent individual users to the model. Are there aspects of an individual's belief system that a language model can utilize for improved alignment? Following prior research, we investigate this question in the domain of opinion prediction by developing PrimeX, a dataset of public opinion survey data from 858 US residents with two additional sources of belief information: written explanations from the respondents for why they hold specific opinions, and the Primal World Belief survey for assessing respondent worldview. We provide an extensive initial analysis of our data and show the value of belief explanations and worldview for personalizing language models. Our results demonstrate how the additional belief information in PrimeX can benefit both the NLP and psychological research communities, opening up avenues for further study.


Sometimes the Model doth Preach: Quantifying Religious Bias in Open LLMs through Demographic Analysis in Asian Nations

Shankar, Hari, P, Vedanta S, Cavale, Tejas, Kumaraguru, Ponnurangam, Chakraborty, Abhijnan

arXiv.org Artificial Intelligence

Large Language Models (LLMs) are capable of generating opinions and propagating bias unknowingly, originating from unrepresentative and non-diverse data collection. Prior research has analysed these opinions with respect to the West, particularly the United States. However, insights thus produced may not be generalized in non-Western populations. With the widespread usage of LLM systems by users across several different walks of life, the cultural sensitivity of each generated output is of crucial interest. Our work proposes a novel method that quantitatively analyzes the opinions generated by LLMs, improving on previous work with regards to extracting the social demographics of the models. Our method measures the distance from an LLM's response to survey respondents, through Hamming Distance, to infer the demographic characteristics reflected in the model's outputs. We evaluate modern, open LLMs such as Llama and Mistral on surveys conducted in various global south countries, with a focus on India and other Asian nations, specifically assessing the model's performance on surveys related to religious tolerance and identity. Our analysis reveals that most open LLMs match a single homogeneous profile, varying across different countries/territories, which in turn raises questions about the risks of LLMs promoting a hegemonic worldview, and undermining perspectives of different minorities. Our framework may also be useful for future research investigating the complex intersection between training data, model architecture, and the resulting biases reflected in LLM outputs, particularly concerning sensitive topics like religious tolerance and identity.


Mapping the Podcast Ecosystem with the Structured Podcast Research Corpus

Litterer, Benjamin, Jurgens, David, Card, Dallas

arXiv.org Artificial Intelligence

Podcasts provide highly diverse content to a massive listener base through a unique on-demand modality. However, limited data has prevented large-scale computational analysis of the podcast ecosystem. To fill this gap, we introduce a massive dataset of over 1.1M podcast transcripts that is largely comprehensive of all English language podcasts available through public RSS feeds from May and June of 2020. This data is not limited to text, but rather includes audio features and speaker turns for a subset of 370K episodes, and speaker role inferences and other metadata for all 1.1M episodes. Using this data, we also conduct a foundational investigation into the content, structure, and responsiveness of this ecosystem. Together, our data and analyses open the door to continued computational research of this popular and impactful medium.


Hidden Persuaders: LLMs' Political Leaning and Their Influence on Voters

Potter, Yujin, Lai, Shiyang, Kim, Junsol, Evans, James, Song, Dawn

arXiv.org Artificial Intelligence

How could LLMs influence our democracy? We investigate LLMs' political leanings and the potential influence of LLMs on voters by conducting multiple experiments in a U.S. presidential election context. Through a voting simulation, we first demonstrate 18 open- and closed-weight LLMs' political preference for a Democratic nominee over a Republican nominee. We show how this leaning towards the Democratic nominee becomes more pronounced in instruction-tuned models compared to their base versions by analyzing their responses to candidate-policy related questions. We further explore the potential impact of LLMs on voter choice by conducting an experiment with 935 U.S. registered voters. During the experiments, participants interacted with LLMs (Claude-3, Llama-3, and GPT-4) over five exchanges. The experiment results show a shift in voter choices towards the Democratic nominee following LLM interaction, widening the voting margin from 0.7% to 4.6%, even though LLMs were not asked to persuade users to support the Democratic nominee during the discourse. This effect is larger than many previous studies on the persuasiveness of political campaigns, which have shown minimal effects in presidential elections. Many users also expressed a desire for further political interaction with LLMs. Which aspects of LLM interactions drove these shifts in voter choice requires further study. Lastly, we explore how a safety method can make LLMs more politically neutral, while raising the question of whether such neutrality is truly the path forward.


Are Social Sentiments Inherent in LLMs? An Empirical Study on Extraction of Inter-demographic Sentiments

Tanaka, Kunitomo, Sasano, Ryohei, Takeda, Koichi

arXiv.org Artificial Intelligence

Large language models (LLMs) are supposed to acquire unconscious human knowledge and feelings, such as social common sense and biases, by training models from large amounts of text. However, it is not clear how much the sentiments of specific social groups can be captured in various LLMs. In this study, we focus on social groups defined in terms of nationality, religion, and race/ethnicity, and validate the extent to which sentiments between social groups can be captured in and extracted from LLMs. Specifically, we input questions regarding sentiments from one group to another into LLMs, apply sentiment analysis to the responses, and compare the results with social surveys. The validation results using five representative LLMs showed higher correlations with relatively small p-values for nationalities and religions, whose number of data points were relatively large. This result indicates that the LLM responses including the inter-group sentiments align well with actual social survey results.


Americans growing anxious as AI adoption expands, Pew Research finds

Engadget

Americans have grown more worried about AI in the last nine months. A new survey from the Pew Research Center indicates 52 percent of respondents are more concerned than excited about rising artificial intelligence use, up 14 points since December. Meanwhile, only 10 percent say they're more excited than worried, while another 36 percent described their views as equally balanced. "Concern about AI outweighs excitement across all major demographic groups," the Pew Research Center wrote in a blog post today. It's been an eventful nine months since the Pew Center last surveyed people about AI.


The Future of Human Agency

#artificialintelligence

This report covers results from the 15th "Future of the Internet" canvassing that Pew Research Center and Elon University's Imagining the Internet Center have conducted together to gather expert views about important digital issues. This is a nonscientific canvassing based on a nonrandom sample; this broad array of opinions about the potential influence of current trends may lead between 2022 and 2035 represents only the points of view of the individuals who responded to the queries. Pew Research Center and Elon's Imagining the Internet Center sampled from a database of experts to canvass from a wide range of fields, inviting entrepreneurs, professionals and policy people based in government bodies, nonprofits and foundations, technology businesses and think tanks, as well as interested academics and technology innovators. The predictions reported here came in response to a set of questions in an online canvassing conducted between June 29 and Aug. 8, 2022. In all, 540 technology innovators and developers, business and policy leaders, researchers and activists responded in some way to the question covered in this report. More on the methodology underlying this canvassing and the participants can be found in the section titled "About this canvassing of experts." Advances in the internet, artificial intelligence (AI) and online applications have allowed humans to vastly expand their capabilities and increase their capacity to tackle complex problems. These advances have given people the ability to instantly access and share knowledge and amplified their personal and collective power to understand and shape their surroundings. Today there is general agreement that smart machines, bots and systems powered mostly by machine learning and artificial intelligence will quickly increase in speed and sophistication between now and 2035.


Meet the Speaker -- Alley Lyles-Jenkins

#artificialintelligence

Why is speaking at WITS important to you? WITS events have been enjoyable for me as a speaker and attendee. I filled my schedule with panels related to technology's business and technical aspects. I appreciate that the programming is diverse in its approach to the topics covered under the Tech umbrella. What inspires me: The advancements associated with DALL*E, the AI system that creates realistic images and art from natural language descriptions, are astounding. There are implications for content, advertising, and investment opportunities.


Using Data To Combat COVID-19

#artificialintelligence

Let's examine the communication breakdown from a more human perspective: Reporting numbers and statistics is only half the battle. The person receiving the information has to decode, interpret, and decide for themselves what this information means to them. This includes exacerbating factors like confirmation bias, and the inaccessibility of credible information (or, put another way, the relative ease with which one can find misinformation). Confirmation bias is a type of cognitive bias where information that is consistent with our existing beliefs is weighted with greater importance and is more likely to be remembered than information that is inconsistent with our existing beliefs. For example, if you prefer the colour red to blue, and there are equally valid studies published on why red is better and why blue is better, you're more likely to believe and remember the study that supports your love of the colour red.